Actor-critic continuous state reinforcement learning for wind-turbine control robust optimization
نویسندگان
چکیده
The control of Variable-Speed Wind-Turbines (VSWT) extracting electrical power from the wind kinetic energy are composed subsystems that need to be controlled jointly, namely blade pitch and generator torque controllers. Previous state art approaches decompose joint problem into independent subproblems, each with its own subgoal, carrying out separately design tuning a parameterized controller for subproblem. Such neglect interactions among which can introduce significant effects. This paper applies Actor-Critic Reinforcement Learning (ACRL) as whole, simultaneous parameter optimization both without neglecting their interactions, aiming globally optimal whole system. innovative architecture uses an augmented input space so parameters fine-tuned working condition. Validation results conducted on simulation experiments using state-of-the-art OpenFAST simulator show efficiency improvement relative best controllers used benchmarks, up 22% in average error performance after ACRL training.
منابع مشابه
Dynamic Control with Actor-Critic Reinforcement Learning
4 Actor-Critic Marble Control 4 4.1 R-code . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 4.2 The critic . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 4.3 Unstable actors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 4.4 Trading off stability against...
متن کاملActor-Critic Reinforcement Learning with Neural Networks in Continuous Games
Reinforcement learning agents with artificial neural networks have previously been shown to acquire human level dexterity in discrete video game environments where only the current state of the game and a reward are given at each time step. A harder problem than discrete environments is posed by continuous environments where the states, observations, and actions are continuous, which is what th...
متن کاملActor-Critic Reinforcement Learning with Simultaneous Human Control and Feedback
This paper contributes a first study into how different human users deliver simultaneous control and feedback signals during human-robot interaction. As part of this work, we formalize and present a general interactive learning framework for online cooperation between humans and reinforcement learning agents. In many humanmachine interaction settings, there is a growing gap between the degrees-...
متن کامل1 Supervised Actor - Critic Reinforcement Learning
Editor’s Summary: Chapter ?? introduced policy gradients as a way to improve on stochastic search of the policy space when learning. This chapter presents supervised actor-critic reinforcement learning as another method for improving the effectiveness of learning. With this approach, a supervisor adds structure to a learning problem and supervised learning makes that structure part of an actor-...
متن کاملDiff-DAC: Distributed Actor-Critic for Multitask Deep Reinforcement Learning
We propose a multiagent distributed actor-critic algorithm for multitask reinforcement learning (MRL), named Diff-DAC. The agents are connected, forming a (possibly sparse) network. Each agent is assigned a task and has access to data from this local task only. During the learning process, the agents are able to communicate some parameters to their neighbors. Since the agents incorporate their ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Information Sciences
سال: 2022
ISSN: ['0020-0255', '1872-6291']
DOI: https://doi.org/10.1016/j.ins.2022.01.047